Learning Priors for Invariance

نویسندگان

  • Eric Nalisnick
  • Padhraic Smyth
چکیده

For modern large-scale Bayesian models, informative priors are difficult if not impossible to elicit. Yet, often, some prior knowledge is known, and this information is incorporated via engineering tricks or methods less principled than a Bayesian prior. However, employing these tricks is difficult to reconcile with principled probabilistic inference. For instance, in the case of data set augmentation, the posterior is conditioned on artificial data and not what is truly observed. In this paper, we address the problem of how to specify an informative prior when the problem of interest is known to exhibit some invariance properties. The proposed method is akin to posterior variational inference: we choose a parametric family and optimize to find the member of the family that makes the model robust to a given transformation. We demonstrate the method’s utility for dropout and rotation transformations, showing that the use of these priors results in performance competitive to that of non-Bayesian methods. Furthermore, our approach does not depend on the data being labeled and thus can be used in semisupervised settings.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Sparse, complex-valued representations of natural sounds learned with phase and amplitude continuity priors

Complex-valued sparse coding is a data representation which employs a dictionary of two-dimensional subspaces, while imposing sparse, factorial prior on complex amplitudes. When trained on a dataset of natural image patches, it learns phase invariant features which closely resemble receptive fields of complex cells in the visual cortex. Features trained on natural sounds however, rarely reveal ...

متن کامل

On the Invariance of Noninformative Priors by Gauri

Jeffreys’ prior, one of the widely used noninformative priors, remains invariant under reparameterization, but does not perform satisfactorily in the presence of nuisance parameters. To overcome this deficiency, recently various noninformative priors have been proposed in the literature. Ž . This article explores the invariance or lack thereof of some of these noninformative priors including th...

متن کامل

Learning Invariant Representations of Molecules for Atomization Energy Prediction

The accurate prediction of molecular energetics in chemical compound space is a crucial ingredient for rational compound design. The inherently graph-like, non-vectorial nature of molecular data gives rise to a unique and difficult machine learning problem. In this paper, we adopt a learning-from-scratch approach where quantum-mechanical molecular energies are predicted directly from the raw mo...

متن کامل

Mixtures of g-Priors in Generalized Linear Models

Mixtures of Zellner’s g-priors have been studied extensively in linear models and have been shown to have numerous desirable properties for Bayesian variable selection and model averaging. Several extensions of g-priors to Generalized Linear Models (GLMs) have been proposed in the literature; however, the choice of prior distribution of g and resulting properties for inference have received con...

متن کامل

Propriety of Intrinsic Priors in Invariant Testing Situations

The Theory of Intrinsic Priors, developed by Berger and Pericchi (1996a,b), is a general method of constructing objective priors for testing and model selection when proper priors are considered for the simpler null hypotheses. When this prior distribution is improper, as is typically the case for Objective Bayesian testing, they suggest approximating the (improper) prior by a sequence of prope...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2018